92 research outputs found

    Data Mining in Healthcare: A Survey of Techniques and Algorithms with its Limitations and Challenges

    Get PDF
    The large amount of data in healthcare industry is a key resource to be processed and analyzed for knowledge extraction. The knowledge discovery is the process of making low-level data into high-level knowledge. Data mining is a core component of the KDD process. Data mining techniques are used in healthcare management which improve the quality and decrease the cost of healthcare services. Data mining algorithms are needed in almost every step in KDD process ranging from domain understanding to knowledge evaluation. It is necessary to identify and evaluate the most common data mining algorithms implemented in modern healthcare services. The need is for algorithms with very high accuracy as medical diagnosis is considered as a significant yet obscure task that needs to be carried out precisely and efficiently

    Reverse Engineering: Methodologies for Web Applications

    Get PDF
    The Reverse Engineering of Web Applications is a complex problem, due to the variety of languages and technologies that are contemporary used to realize them. Indeed, the benefits that can be obtained are remarkable: the presence of documentation at different abstraction levels will help the execution of maintenance interventions, migration and reengineering processes, reducing their costs and risks and improving their effectiveness. Moreover, the assessment of the maintainability factor of a Web Application is an important support to decision making processes. Business processes are often implemented by mean of software systems which expose them to the user as an externally accessible Web application. This paper describes a methodologies for recovering business processes by dynamic analysis of the Web applications which ex-pose them

    Cloud based intrusion detection architecture for smartphones

    Get PDF
    Smartphones are phones with advanced capabilities like those of personal computers (PCs). Smartphone technology is more and more becoming the predominant communication tool for people across the world. People use their smartphones to keep their contact data, to browse the internet, to exchange messages, to keep notes, carry their personal files and documents, etc. Users while browsing are also capable of shopping online, thus provoking a need to type their credit card numbers and security codes. As the smartphones are becoming widespread, it's also becoming a popular target for security threats and attack. Since smartphones use the same software architecture as in PCs, they are vulnerable to be exposed to similar threats such as in PCs. Recent news and articles indicate huge increase in malware and viruses for operating systems employed on smartphones (primarily Android and iOS). Major limitations of smartphone technology are its processing power and its scarce energy source since smartphones rely on battery usage. The smartphones have less storage and computational power to put into effect highly complex algorithms for intrusion detection and implementing signature based attack detection. Now in this paper, we propose a cloud based Intrusion Detection System for smartphones to overcome the issues of smartphone resource constraints and to detect any misbehavior or anomalous activity effectively

    INDIAN LANGUAGE TEXT MINING

    Get PDF
    India is the home of different languages, due to its cultural and geographical diversity. In the Constitution of India, a provision is made for each of the Indian states to choose their own official language for communicating at the state level for official purpose. In India, the growth in consumption of Indian language content started because of growth of electronic devices and technology. The availability of constantly increasing amount of textual data of various Indian regional languages in electronic form has accelerated. But not much work has been done in Indian languages text processing. So there is a huge gap from the stored data to the knowledge that could be constructed from the data. This transition won't occur automatically, that's where Text mining comes into picture. This research is concerned with the study and analyzes the text mining for Indian regional languages Text mining refers to such a knowledge discovery process when the source data under consideration is text. Text mining is a new and exciting research area that tries to solve the information overload problem by using techniques from information retrieval, information extraction as well as natural language processing (NLP) and connects them with the algorithms and methods of KDD, data mining, machine learning and statistics. Some applications of text mining are: document classification, information retrieval, clustering documents, information extraction, and performance evaluation. In this paper we made an attempt to show the need of text mining for Indian language

    Interlingual Machine Translation

    Get PDF
    Interlingual is an artificial language used to represent the meaning of natural languages, as for purposes of machine translation. It is an intermediate form between two or more languages. Machine translation is the process of translating from source language text into the target language. This paper proposes a new model of machine translation system in which rule-based and example-based approaches are applied for English-to-Kannada/Telugu sentence translation. The proposed method has 4 steps: 1) analyze an English sentence into a string of grammatical nodes, based on Phrase Structure Grammar, 2) map the input pattern with a table of English-Kannada/Telugu sentence patterns, 3) look up the bilingual dictionary for the equivalent Kannada/Telugu words, reorder and then generate output sentences and 4) rank the possible combinations and eliminate the ambiguous output sentences by using a statistical method. The translated sentences will then be stored in a bilingual corpus to serve as a guide or template for imitating the translation, i.e., the example-based approach. The future work will focus on sentence translation by using semantic features to make a more precise translation

    In Vitro Techniques to Accelerate Flavonoid Synthesis in some Euphorbiaceae Members

    Get PDF
    Intrusion detection is the process of monitoring the events occurring in a computer system or network and analyzing them for signs of possible incidents, which are violations or imminent threats of violation of computer security policies, acceptable use policies, or standard security practices. An intrusion detection system (IDS) monitors network traffic and monitors for suspicious activity and alerts the system or network administrator. It identifies unauthorized use, misuse, and abuse of computer systems by both system insiders and external penetrators. Intrusion detection systems (IDS) are essential components in a secure network environment, allowing for early detection of malicious activities and attacks. By employing information provided by IDS, it is possible to apply appropriate countermeasures and mitigate attacks that would otherwise seriously undermine network security. However, current high volumes of network traffic overwhelm most IDS techniques requiring new approaches that are able to handle huge volume of log and packet analysis while still maintaining high throughput. Hadoop, an open-source computing platform of MapReduce and a distributed file system, has become a popular infrastructure for massive data analytics because it facilitates scalable data processing and storage services on a distributed computing system consisting of commodity hardware. The proposed architecture is able to efficiently handle large volumes of collected data and consequent high processing loads using Hadoop, MapReduce and cloud computing infrastructure. The main focus of the paper is to enhance the throughput and scalability of the IDS Log analysi

    Kannada and Telugu Native Languages to English Cross Language Information Retrieval

    Get PDF
    One of the crucial challenges in cross lingual information retrieval is the retrieval of relevant information for a query expressed in as native language. While retrieval of relevant documents is slightly easier, analysing the relevance of the retrieved documents and the presentation of the results to the users are non-trivial tasks. To accomplish the above task, we present our Kannada English and Telugu English CLIR systems as part of Ad-Hoc Bilingual task. We take a query translation based approach using bi-lingual dictionaries. When a query words not found in the dictionary then the words are transliterated using a simple rule based approach which utilizes the corpus to return the ‘k’ closest English transliterations of the given Kannada/Telugu word. The resulting multiple translation/transliteration choices for each query word are disambiguated using an iterative page-rank style algorithm which, based on term-term co-occurrence statistics, produces the final translated query. Finally we conduct experiments on these translated query using a Kannada/Telugu document collection and a set of English queries to report the improvements, performance achieved for each task is to be presented and statistical analysis of these results are given

    Effect of Feature Selection to Improve Accuracy and Decrease Execution Time with Predicating Learning Disabilities in School Going Children

    Get PDF
    Learning disability in school children is the representation of brain disorder which includes several disorders in which school going child faces the difficulties. The evaluation of learning disability is a crucial and important task in the field of educational field. This process can be accomplished by using data mining approaches. The efficiency of this approach is based on the feature selection while performing the prediction of the learning disabilities. In paper mainly aims on the efficient method of feature selection to improve the accuracy of prediction and classification in school going children. Feature selection is a process to collect the small subset of the features from huge dataset. A commonly used approach in feature selection is ranking the individual features according to some criteria and then search for an optimal feature subset based on evaluation criterion to test the optimality. In the Wrapper model we use some predetermined learning algorithm to find out the relevant features and test them. It requires more computations, so if there are large numbers of features we prefer to filter. In this paper first we have used feature selection attribute algorithms Chi-square. Info Gain, and Gain Ratio to predict the relevant features. Then we have applied fast correlation base filter algorithm on given features. Later classification is done using KNN and SVM. Results showed reduction in computational cost and time and increase in predictive accuracy for the student model. The objective of this work is to predict the presence of Learning Disability (LD) in school-aged children more accurately and help them to develop a bright future according to his choice by predicting the success at the earliest

    Nlp Challenges for Machine Translation from English to Indian Languages

    Get PDF
    This Natural Langauge processing is carried particularly on English-Kannada/Telugu. Kannada is a language of India. The Kannada language has a classification of Dravidian, Southern, Tamil-Kannada, and Kannada. Regions Spoken: Kannada is also spoken in Karnataka, Andhra Pradesh, Tamil Nadu, and Maharashtra. Population: The total population of people who speak Kannada is 35,346,000, as of 1997. Alternate Name: Other names for Kannada are Kanarese, Canarese, Banglori, and Madrassi. Dialects: Some dialects of Kannada are Bijapur, Jeinu Kuruba, and Aine Kuruba. There are about 20 dialects and Badaga may be one. Kannada is the state language of Karnataka. About 9,000,000 people speak Kannada as a second language. The literacy rate for people who speak Kannada as a first language is about 60%, which is the same for those who speak Kannada as a second language (in India). Kannada was used in the Bible from 1831-2000. Statistical machine translation (SMT) is a machine translation paradigm where translations are generated on the basis of statistical models whose parameters are derived from the analysis of bilingual text corpora. The statistical approach contrasts with the rule-based approaches to machine translation as well as with example-based machine translatio

    Hybrid classification approach hdlmm for learning disability prediction in school going children using data mining technique

    Get PDF
    Learning Disability is a disorder of neurological condition which causes deficiency in child�s brain activities such as reading, speaking and many other tasks. According to the World Health Organization (WHO), 15 of the children get affected by the learning disability. Efficient prediction and accurate classification is the crucial task for researchers for early detection of learning disability. In this work, our main aim to develop a model for learning disability prediction and classification with the help of soft computing technique. To improve the performance of the prediction and classification we propose a hybrid approach for feature reduction and classification. Proposed approach is divided into three main stages: (i) data pre-processing (ii) feature selection and reduction and (iii) Classification. In this approach, preprocessing, feature selection and reduction is carried out by measuring of confidence with adaptive genetic algorithm. Prediction and classification is carried out by using Deep Learner Neural network and Markov Model. Genetic algorithm is used for data preprocessing to achieve the feature reduction and confidence measurement. The system is implemented using MatLab 2013b. Result analysis shows that the proposed approach is capable to predict the learning disability effectively. © 2005 � ongoing JATIT & LLS
    • …
    corecore